Grid Computing in the Collider Detector at Fermilab (CDF) scientific experiment

نویسنده

  • Douglas P. Benjamin
چکیده

CDF detector measures the particles produced in proton – anti proton collisions at the Tevatron operating at s = 1.96 TeV The are two experiments at the Tevatron and this paper will describe the grid computing for CDF. The CDF detector is comprised of a tracking system located closest to the collision point inside of a solenoid magnet. Outside of the solenoid magnet, calorimeter detectors used to measure the energy of the particles resulting from the collisions. Finally on the outside are muon detectors. The CDF experiment started collecting data in 1988 and will continue to take data until at least October 2009. There are 3⁄4 million electronic channels readout and there are collisions every 396 ns. The CDF computing model has several components. The events passing the trigger are readout and written to robotic tape storage for further calibration, reconstruction and analysis. The reconstruction program reads the data from tape using dCache system [1] to stage the data to disk prior to copying the data files to the worker nodes located in an onsite computer farm and writing the results to tape. CDF users typically analyze the data at FNAL and use remote computer farms for Monte Carlo production with the results send back to FNAL for archival storage. The CDF analysis farm (CAF) [2] is a software infrastructure developed to allow CDF to use computer farms of commodity hardware for data reconstruction, data analysis, simulated data production. Users develop software, debug programs and submit jobs from their own desktops. User authentication is based on Kerberos v5. Monitoring of system performance and user jobs can be done either through a web interface or command line tools. The interactive monitoring tools allow the users to check the progress of their running jobs by viewing the log files on the worker nodes and obtain other information from the worker nodes (like directory of user job’s remote working area). Users can also stop all of part of their jobs if needed. Most of the monitoring functionality is available to the CDF users running on the grid. The users are notified of the completion of their job via job summary sent via e-mail. The user can also specify the location to which output compressed tar ball containing the job logs and whatever files existed at the job completion in the users work area. Kerberized rcp is used to transfer the files at present. The current version of the CDF CAF used for processing the data with the production reconstruction executable at FNAL is based on a dedicated Condor pool [3]. Condor daemons (collector, negotiator, multiple schedd) run on the head node (or portal machine). The condor starter and startd daemons run on the worker nodes. CAF code includes the job submission client run on the user’s desktop. CAF code used on the portal machine includes job server daemon, monitoring daemons (web and command line), mailer daemons. On the worker nodes a CDF specific job wrapper containing the user job executables is used. The initial implementation was not grid complaint and used remote dedicated computer farms. It has since been adapted to use computer farms located at grid sites. Since the users interact to the CDF computing through portals, the users are removed from the underlying computing farm architecture.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Grid Based Monitoring on the Rutgers CDF Analysis Farm

Run II at the Fermilab Tevatron Collider started in March 2001, and it will continue probing the high energy frontier in particle physics until the start of the LHC at CERN. The CDF collaboration at Fermilab has already stored 260 TB of data and expects to store 1PB of data in the next two years. The HEXCAF computing farm is being set up at Rutgers University to provide the software environment...

متن کامل

Reprocessing D0 Data with SAMGrid

The DØ experiment studies proton-antiproton collisions at the Tevatron collider based at Fermilab. Reprocessing, managing and distributing the large amount of real data coming from the detector as well as generating sufficient Monte Carlo data are some of the challenges faced by the DØ collaboration. SAMGrid combines the SAM data handling system with the necessary job and information management...

متن کامل

Reprocessing DØ data with SAMGrid

The DØ experiment studies proton-antiproton collisions at the Tevatron collider based at Fermilab. Reprocessing, managing and distributing the large amount of real data coming from the detector as well as generating sufficient Monte Carlo data are some of the challenges faced by the DØ collaboration. SAMGrid combines the SAM data handling system with the necessary job and information management...

متن کامل

Diffractive Production of Jets and Vector Bosons at the Tevatron

Recent results on diffractive dijet and vector boson production and exclusive dijet production from the Collider Detector at Fermilab (CDF) experiment are presented.

متن کامل

CDF's Utilisation of Switched Circuit Optical Networks

The ESLEA project utilises the UKLight switched circuit optical network to enable guaranteed high bandwidth network links for a range of eScience applications. The high energy particle physics (HEP) experiment, Collider Detector at Fermilab (CDF), is one such application endeavouring to exploit the potential of high speed optical network links. CDF produces a large amount of raw data (presently...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/0810.3453  شماره 

صفحات  -

تاریخ انتشار 2008